Exercise I: Principal Component Analysis
Recall the mtcars dataset we work with before, which compirses fuel consumption and other aspects of design and performance for 32 cars from 1974. The dataset has 11 dimensions, that is more than it is possible to visualize at the same.
head(mtcars)
- Use
prcomp() to compute a PCA for mtcars. Remember to set the scale parameter, as the variables are in different units and have different ranges
mtcars.pca <- prcomp(mtcars, scale=TRUE)
- Generate a scree plot and note how many dimensions should you retain.
library(factoextra)
Loading required package: ggplot2
Welcome! Want to learn more? See two factoextra-related books at https://goo.gl/ve3WBa
# Percent of variance explained:
fviz_eig(mtcars.pca)
Registered S3 method overwritten by 'data.table':
method from
print.data.table

- Compute the percentage of variance explained by each of the principal components.
eig <- mtcars.pca$sdev^2
(var.exp <- 100*eig/sum(eig))
[1] 60.0763659 24.0951627 5.7017934 2.4508858 2.0313737 1.9236011 1.2296544 1.1172858 0.7004241 0.4730495
[11] 0.2004037
- Generate a biplot for the PCA projection. Use the loadings matrix to inspect which variables contributes most to PC1 and which to PC2. What do the PC1 and PC2 correspond to? How are the cars distributed on this representation? Does the "car map" make sense?
fviz_pca_biplot(mtcars.pca) + coord_fixed()

mtcars.pca$rotation
PC1 PC2 PC3 PC4 PC5 PC6 PC7 PC8 PC9
mpg -0.3625305 0.01612440 -0.22574419 -0.022540255 0.10284468 -0.10879743 0.367723810 -0.754091423 0.235701617
cyl 0.3739160 0.04374371 -0.17531118 -0.002591838 0.05848381 0.16855369 0.057277736 -0.230824925 0.054035270
disp 0.3681852 -0.04932413 -0.06148414 0.256607885 0.39399530 -0.33616451 0.214303077 0.001142134 0.198427848
hp 0.3300569 0.24878402 0.14001476 -0.067676157 0.54004744 0.07143563 -0.001495989 -0.222358441 -0.575830072
drat -0.2941514 0.27469408 0.16118879 0.854828743 0.07732727 0.24449705 0.021119857 0.032193501 -0.046901228
wt 0.3461033 -0.14303825 0.34181851 0.245899314 -0.07502912 -0.46493964 -0.020668302 -0.008571929 0.359498251
qsec -0.2004563 -0.46337482 0.40316904 0.068076532 -0.16466591 -0.33048032 0.050010522 -0.231840021 -0.528377185
vs -0.3065113 -0.23164699 0.42881517 -0.214848616 0.59953955 0.19401702 -0.265780836 0.025935128 0.358582624
am -0.2349429 0.42941765 -0.20576657 -0.030462908 0.08978128 -0.57081745 -0.587305101 -0.059746952 -0.047403982
gear -0.2069162 0.46234863 0.28977993 -0.264690521 0.04832960 -0.24356284 0.605097617 0.336150240 -0.001735039
carb 0.2140177 0.41357106 0.52854459 -0.126789179 -0.36131875 0.18352168 -0.174603192 -0.395629107 0.170640677
PC10 PC11
mpg 0.13928524 -0.124895628
cyl -0.84641949 -0.140695441
disp 0.04937979 0.660606481
hp 0.24782351 -0.256492062
drat -0.10149369 -0.039530246
wt 0.09439426 -0.567448697
qsec -0.27067295 0.181361780
vs -0.15903909 0.008414634
am -0.17778541 0.029823537
gear -0.21382515 -0.053507085
carb 0.07225950 0.319594676
fviz_contrib(mtcars.pca, choice = "var", axes = 1)

fviz_contrib(mtcars.pca, choice = "var", axes = 2)

Exercise 2: Cluster Analysis
Part 1: k-means clustering
We will generate synthetic clustered data to use for k-means clustering.
set.seed(489576)
N <- 1000
C1 <- data.frame(cluster = "C1", x = rnorm(n = N, mean = 1), y = rnorm(n = N, mean = 1))
C2 <- data.frame(cluster = "C2", x = rnorm(n = N, mean = -2), y = rnorm(n = N, mean = -5))
C3 <- data.frame(cluster = "C3", x = rnorm(n = N, mean = 5), y = rnorm(n = N, mean = 1))
DF <- rbind(C1, C2, C3)
ggplot(DF, aes(x, y, color = cluster)) +
geom_point()

- Apply k-means with k = 3 (as you know the true number of clusters). Pring the cluster centers.
kmeans.res <- kmeans(x = DF[, -1], centers = 3)
kmeans.res$centers
x y
1 0.9614141 1.0058486
2 -2.0431190 -4.9932291
3 5.0098346 0.9780214
- Print a confusion map to compare k-means cluster assignment with the true cluster labels.
table(kmeans = kmeans.res$cluster, true = DF$cluster)
true
kmeans C1 C2 C3
1 968 1 23
2 0 999 0
3 32 0 977
- Generate a scatter plot of points, now colored by the cluster assignment.
library(ggplot2)
DF$kmeans <- factor(kmeans.res$cluster)
ggplot(DF, aes(x, y)) +
geom_point(alpha = 0.5, aes( color = kmeans)) +
geom_point(data = data.frame(x = kmeans.res$centers[, 1],
y = kmeans.res$centers[, 2]), size = 3, aes(x, y), color = "Black")

- Now pretend that you don't know the real number of clusters. Use k = 4 and recompute kmeans. Plot the results and see what happened.
kmeans.res2 <- kmeans(x = DF[, -1], centers = 4)
kmeans.res2$centers
x y kmeans
1 5.3301476 1.7351006 3
2 4.7001362 0.2460306 3
3 0.9614141 1.0058486 1
4 -2.0431190 -4.9932291 2
DF$kmeans2 <- factor(kmeans.res2$cluster)
ggplot(DF, aes(x, y, color = kmeans2)) +
geom_point(alpha = 0.5)

Part 2: Hierarchical Clustering
In this exercise you will you use a dataset published in a study by Khan et al. 2001 to perform a hierarchical clustering of the patients in the study based on their overall gene expression data.
This data set consists of expression levels for 2,308 genes. The training and test sets consist of 63 and 20 observations (tissue samples) respectively.
Here, we will use the train set, as we now are only interested in learning how hclust() works. First, load the ISLR where the data is available. The gene expression data is available in an object Khan$xtrain; you can learn more about the data set by typing in ?Khan after loading ISLR package.
library(ISLR)
gene.expression <- Khan$xtrain
dim(gene.expression)
[1] 63 2308
- Compute a (Euclidean) distance matrix between each pair of samples.
D <- dist(gene.expression)
- Perform hierarchical clustering using average linkage.
khan.hclust <- hclust(D, method = "average")
- Plot a dendrogram associated with the hierarchical clustering you just computed. In this example, you actually have the lables of the tissue samples, however, the algorithms was blinded to them. By adding labels to the dendrogram corresponding to
Khan$ytrain, check if the clustering performed groups the observations from same tumor class nearby.
plot(khan.hclust, labels = Khan$ytrain)

Exercise 4: Ads
- Read the data from "https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv" containing information on sales of a product and the amount spent on advertising using different media channels.
url <- "https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv"
sales <- read.csv(url)
cannot open URL 'https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv': HTTP status was '404 Not Found'Error in file(file, "rt") :
cannot open the connection to 'https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv'
- Generate a scatterplot of sales against the amount of TV advertising and add a linear fit line.
ggplot(sales, aes(x = TV, y = Sales)) +
geom_point(aes(fill = Radio), color = "grey", pch = 21, size = 3) +
geom_smooth(method = "lm") +
scale_fill_viridis_c() +
theme_bw()

- Now make a 3D scatterplot with axes corresponding to 'sales', 'TV' and 'radio'.
library(plotly)
Registered S3 method overwritten by 'htmlwidgets':
method from
print.htmlwidget tools:rstudio
Attaching package: ‘plotly’
The following object is masked from ‘package:ggplot2’:
last_plot
The following object is masked from ‘package:stats’:
filter
The following object is masked from ‘package:graphics’:
layout
plot_ly(x=sales$Sales, y=sales$TV, z=sales$Radio, type="scatter3d", mode="markers")
`arrange_()` is deprecated as of dplyr 0.7.0.
Please use `arrange()` instead.
See vignette('programming') for more help
[90mThis warning is displayed once every 8 hours.[39m
[90mCall `lifecycle::last_warnings()` to see where this warning was generated.[39m
- The dataset has 200 rows. Divide it into a train set with 150 observations and a test set with 50 observations, i.e. use
sample(1:200, n = 150) to randomly choose row indices of the advertising dataset to include in the train set. The remaining indices should be used for the test set. Remember to choose and set the seed for randomization!
set.seed(123)
idx <- sample(1:200, size = 150)
train <- sales[idx, ]
test <- sales[-idx, ]
fit <- lm(Sales ~ TV, train)
pred_test <- predict(fit, test)
RMSE <- sqrt(mean((pred_test - test$Sales)^2))
cat("The RMSE for the simple regression model is:", RMSE, ".\n")
The RMSE for the simple regression model is: 3.43658 .
Fit a linear model to the training set, where the sales values are predicted by the amount of TV advertising. Print the summary of the fitted model. Then, predict the sales values for the test set and evaluate the test model accuracy in terms of root mean squared error (MSE), which measures the average level of error between the prediction and the true response. \[RMSE = \sqrt{\frac{1}{n} \sum\limits_{i = 1}^n(\hat y_i - y_i)^2}\]
Fit a multiple linerar regression model including all the variables 'TV', 'radio', 'newspaper' to model the 'sales' in the training set. Then, compute the predicted sales for the test set with the new model and evalued the RMSE.
Did the error decrease from the one correspodning to the previous model?
fit.mult <- lm(Sales ~ TV + Radio + Newspaper, train)
pred_test <- predict(fit.mult, test)
RMSE <- sqrt(mean((pred_test - test$Sales)^2))
cat("The RMSE for the multiple regression model is:", RMSE, ".\n")
The RMSE for the multiple regression model is: 1.909814 .
Yes, the RMSE decrease after including more predictors in the model.
- Look at the summary output for the multiple regression model and note which of the coefficient in the model is significant. Are all of them significant? If not refit the model including only the features found significant. Which of the models should you choose?
summary(fit.mult)
Call:
lm(formula = Sales ~ TV + Radio + Newspaper, data = train)
Residuals:
Min 1Q Median 3Q Max
-5.3581 -0.8916 0.2041 1.2089 2.6565
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 3.273848 0.339641 9.639 <2e-16 ***
TV 0.044485 0.001544 28.816 <2e-16 ***
Radio 0.190370 0.009643 19.741 <2e-16 ***
Newspaper -0.005491 0.006310 -0.870 0.386
---
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
Residual standard error: 1.612 on 146 degrees of freedom
Multiple R-squared: 0.9038, Adjusted R-squared: 0.9018
F-statistic: 457.1 on 3 and 146 DF, p-value: < 2.2e-16
It seems like the 'newspaper' predictor is not significant, so we will remove it.
fit.mult2 <- lm(Sales ~ TV + Radio, train)
pred_test <- predict(fit.mult2, test)
RMSE <- sqrt(mean((pred_test - test$Sales)^2))
cat("The RMSE for the new multiple regression model is:", RMSE, ".\n")
The RMSE for the new multiple regression model is: 1.890502 .
The second multiple regression model has a slightly better test error. We should not use more predictors than necessary, so 'newspaper' should be discarded from the model.
Doctor Visits
Data was collected on doctor visits from a sample of 5,190 people in the 1977/1978 Australian Health Survey. Cameron (1986) sought to explain the variation in doctor visits using one or more explanatory variables. The data can be found in an R data set from library(AER) accessible with the command data("DoctorVisits"). Variable descriptions can be found under help("DoctorVisits")
Explore the use of a zero-inflated model for this data. Begin with a histogram of the number of visits, some EDA, and fitting several models. Summarize your results. Compare your results with a standard Poisson.
library(AER)
data("DoctorVisits")
ggplot(DoctorVisits) +
geom_histogram(aes(x=visits))+
theme_bw()

ggplot(DoctorVisits) +
geom_histogram(aes(x=visits))+
facet_grid(gender~private)+
theme_bw()

df= DoctorVisits %>% group_by(visits) %>% summarise_if(is.numeric, mean)
Error in DoctorVisits %>% group_by(visits) %>% summarise_if(is.numeric, :
could not find function "%>%"
There are a lot of zeros --- hence the need for a zero-inflated model.
We begin by fitting a standard model and assessing goodness of fit.
summary(pois.m1 <- glm(formula = visits ~ ., family = poisson,
data = DoctorVisits))
Call:
glm(formula = visits ~ ., family = poisson, data = DoctorVisits)
Deviance Residuals:
Min 1Q Median 3Q Max
-2.9502 -0.6858 -0.5747 -0.4852 5.7055
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) -2.097821 0.101554 -20.657 < 2e-16 ***
genderfemale 0.156490 0.056139 2.788 0.00531 **
age 0.279123 0.165981 1.682 0.09264 .
income -0.187416 0.085478 -2.193 0.02834 *
illness 0.186156 0.018263 10.193 < 2e-16 ***
reduced 0.126690 0.005031 25.184 < 2e-16 ***
health 0.030683 0.010074 3.046 0.00232 **
privateyes 0.126498 0.071552 1.768 0.07707 .
freepooryes -0.438462 0.179799 -2.439 0.01474 *
freerepatyes 0.083640 0.092070 0.908 0.36365
nchronicyes 0.117300 0.066545 1.763 0.07795 .
lchronicyes 0.150717 0.082260 1.832 0.06692 .
---
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
(Dispersion parameter for poisson family taken to be 1)
Null deviance: 5634.8 on 5189 degrees of freedom
Residual deviance: 4380.1 on 5178 degrees of freedom
AIC: 6735.7
Number of Fisher Scoring iterations: 6
# Goodness-of-fit test
gof.ts = pois.m1$deviance
gof.pvalue = 1 - pchisq(gof.ts, pois.m1$df.residual)
gof.pvalue
[1] 1
Now we fit the zero-inflated model:
library(pscl)
summary(zi <- zeroinfl(formula = visits ~ health + age + gender | freepoor, data = DoctorVisits))
Call:
zeroinfl(formula = visits ~ health + age + gender | freepoor, data = DoctorVisits)
Pearson residuals:
Min 1Q Median 3Q Max
-0.7366 -0.4761 -0.4119 -0.3744 19.6376
Count model coefficients (poisson with log link):
Estimate Std. Error z value Pr(>|z|)
(Intercept) -1.224332 0.092322 -13.261 < 2e-16 ***
health 0.126664 0.009391 13.488 < 2e-16 ***
age 1.070226 0.143679 7.449 9.42e-14 ***
genderfemale 0.200220 0.060976 3.284 0.00102 **
Zero-inflation model coefficients (binomial with logit link):
Estimate Std. Error z value Pr(>|z|)
(Intercept) 0.14309 0.08059 1.775 0.0758 .
freepooryes 1.12007 0.27857 4.021 5.8e-05 ***
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Number of iterations in BFGS optimization: 12
Log-likelihood: -3585 on 6 Df
vuong(pois.m1,zi)
Vuong Non-Nested Hypothesis Test-Statistic:
(test-statistic is asymptotically distributed N(0,1) under the
null that the models are indistinguishible)
-------------------------------------------------------------
Movies
Recall the movies data-frame we used ealier in the bootcamp. It contains information on movies from the last three decates, which was scrapped from the IMDB database.
library(dplyr)
url <- "https://raw.githubusercontent.com/Juanets/movie-stats/master/movies.csv"
movies <- tbl_df(read.csv(url))
- Generate a boxplot of runtimes for action movies and comedies with jittered points overlaid on top. You might consider setting collor, fill and alpha arguments to modify clarity and transparency of the plot.
library(tidyverse)
ggplot(movies %>% filter(genre %in% c("Action", "Comedy")),aes(x=genre, y = runtime)) +
geom_boxplot(outlier.shape = NA) +
geom_jitter(alpha=0.5, width = 0.1)
- Test a hypothesis that the action movies have higher mean runtime (length) than the comedies. Is the difference statistically greater than zero at significance level \(\alpha = 0.05\)?
comedies = movies %>% filter(genre== "Comedy")
actions = movies %>% filter(genre== "Action")
t.test(comedies$runtime, actions$runtime)
- Test the hypothesis that the scores are the same across movie types (keep the movie genre which have at least 20 movies). Plot the data before making a test of your choice. State all the assumptions that you are making when devising your test.
- The observations are obtained independently and randomly from the population defined by the factor levels
- The data of each factor level are normally distributed.
- These normal populations have a common variance. (Levene’s test can be used to check this.)
d = movies %>%
group_by(genre) %>%
summarise(
count = n(),
mean = mean(score, na.rm = TRUE),
sd = sd(score, na.rm = TRUE)
) %>% filter(count>=20)
genres_selected = d$genre
ggplot(movies %>% filter(genre %in% genres_selected), aes(x=genre, y=score, color=genre)) +
geom_boxplot(outlier.shape = NA) +
geom_jitter(alpha=0.1, width = 0.1) +
theme_bw()
res.aov <- aov(score ~ genre, data =movies %>% filter(genre %in% genres_selected))
# Summary of the analysis
summary(res.aov)
- Is there a reason to believe that the scores might differ according to genre? How would you test which one is different (do the test if you have reason to believe that this is the case).
TukeyHSD(res.aov)
Interpretation:
- diff: difference between means of the two groups
- lwr, upr: the lower and the upper end point of the confidence interval at 95% (default)
- p adj: p-value after adjustment for the multiple comparisons.
- Now, sppose your friend at Hollywood wants to know the recipe for making a movie with the best margin (gross-budget). Which genre and runtime should he aim for? Could you fit a model to try to help him out?
movies$margin = movies$gross - movies$budget
summary(mod <- lm(margin ~ genre+ runtime, data= movies))
---
title: 'Lab 21: In-class: GLMs and Unsupervised Learning'
output:
  html_document:
    df_print: paged
    toc: yes
  html_notebook:
    toc: yes
    toc_float: yes
date: "08/13/2020"
---


# Exercise I: Principal Component Analysis

Recall the `mtcars` dataset we work with before, which compirses fuel 
consumption and other aspects of design and performance for 32 cars from 1974.
The dataset has 11 dimensions, that is more than it is possible to visualize at 
the same.

```{r}
head(mtcars)
```

a. Use `prcomp()` to compute a PCA for `mtcars`. Remember to set the
scale parameter, as the variables are in different units and have different
ranges

```{r}
mtcars.pca <- prcomp(mtcars, scale=TRUE)
```

b. Generate a scree plot and note how many dimensions should you retain.

```{r}
library(factoextra)

# Percent of variance explained:
fviz_eig(mtcars.pca) 
```

c. Compute the percentage of variance explained by each of the principal
components.

```{r}
eig <- mtcars.pca$sdev^2
(var.exp <- 100*eig/sum(eig))
```

d. Generate a biplot for the PCA projection. Use the loadings matrix to inspect
which variables contributes most to PC1 and which to PC2. What do the PC1 and
PC2 correspond to? How are the cars distributed on this representation?
Does the "car map" make sense?

```{r, fig.width=8, fig.height=6}
fviz_pca_biplot(mtcars.pca) + coord_fixed() 
```

```{r}
mtcars.pca$rotation
```

```{r}
fviz_contrib(mtcars.pca, choice = "var", axes = 1) 
```

```{r}
fviz_contrib(mtcars.pca, choice = "var", axes = 2) 
```



# Exercise 2: Cluster Analysis

## Part 1: k-means clustering

We will generate synthetic clustered data to use for k-means clustering.
```{r}
set.seed(489576)
N <- 1000
C1 <- data.frame(cluster = "C1", x = rnorm(n = N, mean = 1), y = rnorm(n = N, mean = 1))
C2 <- data.frame(cluster = "C2", x = rnorm(n = N, mean = -2), y = rnorm(n = N, mean = -5))
C3 <- data.frame(cluster = "C3", x = rnorm(n = N, mean = 5), y = rnorm(n = N, mean = 1))
DF <- rbind(C1, C2, C3)
```

```{r}
ggplot(DF, aes(x, y, color = cluster)) + 
  geom_point()
```

a. Apply k-means with k = 3 (as you know the true number of clusters).
Pring the cluster centers.

```{r}
kmeans.res <- kmeans(x = DF[, -1], centers = 3)
kmeans.res$centers
```

b. Print a confusion map to compare k-means cluster assignment with
the true cluster labels.

```{r}
table(kmeans = kmeans.res$cluster, true = DF$cluster)
```


c. Generate a scatter plot of points, now colored by the cluster assignment.


```{r}
library(ggplot2)
DF$kmeans <- factor(kmeans.res$cluster)
ggplot(DF, aes(x, y)) + 
  geom_point(alpha = 0.5, aes( color = kmeans)) + 
  geom_point(data = data.frame(x = kmeans.res$centers[, 1], 
                               y = kmeans.res$centers[, 2]), size = 3, aes(x, y), color = "Black")
```

d. Now pretend that you don't know the real number of clusters. Use k = 4
and recompute kmeans. Plot the results and see what happened.

```{r}
kmeans.res2 <- kmeans(x = DF[, -1], centers = 4)
kmeans.res2$centers
```

```{r}
DF$kmeans2 <- factor(kmeans.res2$cluster)
ggplot(DF, aes(x, y, color = kmeans2)) + 
  geom_point(alpha = 0.5)
```


## Part 2: Hierarchical Clustering

In this exercise you will you use a dataset published in a study by
[Khan et al. 2001](https://www.nature.com/articles/nm0601_673)
to perform a hierarchical clustering of the patients in the study based
on their overall gene expression data.

This data set consists of expression levels for 2,308 genes.
The training and test sets consist of 63 and 20 observations (tissue samples) 
respectively.

Here, we will use the train set, as we now are only interested in
learning how `hclust()` works. First, load the `ISLR` where the
data is available. The gene expression data is available in an object
`Khan$xtrain`; you can learn more about the data set by typing in `?Khan`
after loading `ISLR` package.

```{r}
library(ISLR)
gene.expression <- Khan$xtrain
dim(gene.expression)
```

a. Compute a (Euclidean) distance matrix between each pair of samples.

```{r}
D <- dist(gene.expression)
```

b. Perform hierarchical clustering using average linkage.

```{r}
khan.hclust <- hclust(D, method = "average")
```

c. Plot a dendrogram associated with the hierarchical clustering you just
computed. In this example, you actually have the lables of the tissue samples,
however, the algorithms was blinded to them. By adding labels to the dendrogram
corresponding to `Khan$ytrain`, check if the clustering performed groups the 
observations from same tumor class nearby. 

```{r}
plot(khan.hclust, labels = Khan$ytrain)
```


## Exercise Extra: 2D visualization of MNIST data

* Download MNIST data of the digits images from 
[Kaggle competition](https://www.kaggle.com/c/digit-recognizer).
* The code is adapted from the one found [here](https://www.kaggle.com/gospursgo/digit-recognizer/clusters-in-2d-with-tsne-vs-pca/code). 

The files are data on the 28x28 pixel
images of digits (0-9). The data is composed of:

* `label` column denoting the digit on the image
* `pixel0` through `pixel783` contain information on the pixel intensity
(on the scale of 0-255), and together form the vectorized version of 
the 28x28 pixel digit image

![](mnistExamples.png)

Download the data from the course repository:

```{r}
# load the already subsetted MNIST data.
mnist.url <- "https://github.com/cme195/cme195.github.io/raw/master/assets/data/mnist_small.csv"
train <- read.csv(mnist.url, row.names = 1)
dim(train)
train[1:10, 1:10]
```

a. Compute and the PCA for the data. Then, extract the first two principal
component scores for the data.

```{r}
# compare with pca
pca <- prcomp(train[,-1])
coord.pca <- data.frame(pca$x[, 1:2])
coord.pca$label <- factor(train$label)
```

b. Plot the 2D principal component scores matrix.

```{r}
ggplot(coord.pca, aes(x= PC1, y = PC2)) + ggtitle("PCA") +
  geom_text(aes(label = label, color = label), alpha = 0.8)
```

c. Compute a tSNE embedding.
```{r}
# Use tsne
library(Rtsne)
set.seed(123) # for reproducibility
tsne <- Rtsne(train[,-1], dims = 2, perplexity=30, 
              verbose=FALSE, max_iter = 500)
coord.tsne <- data.frame(tsne$Y)
coord.tsne$label <- factor(train$label)
```

d. Visualize the tSNE 2D projection.

```{r}
ggplot(coord.tsne, aes(x= X1, y = X2)) + ggtitle("tSNE") +
  geom_text(aes(label = label, color = label), alpha = 0.8)
```

e. What do you observe? How does tSNE compare with PCA in this case?

tSNE seems to be much better at separating digits from each other





# Exercise 4:  Ads

a. Read the data from "https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv"
containing information on sales of a product and the amount spent on advertising
using different media channels.

```{r}
url <- "https://biox-rbootcamp.github.io/biox-rbootcamp.github.io/assets/lectures/unsupervised_learning/Advertising.csv"
sales <- read.csv(url)
sales
```



b. Generate a scatterplot of sales against the amount of TV advertising and 
add a linear fit line.

```{r}
ggplot(sales, aes(x = TV, y = Sales)) +
    geom_point(aes(fill = Radio), color = "grey", pch = 21, size = 3) +
    geom_smooth(method = "lm") +
    scale_fill_viridis_c() +
    theme_bw()
```



c. Now make a 3D scatterplot with axes corresponding to 'sales', 'TV'
and 'radio'.

```{r}
library(plotly)
plot_ly(x=sales$Sales, y=sales$TV, z=sales$Radio, type="scatter3d", mode="markers")
```


d. The dataset has 200 rows. Divide it into a train set with 150 observations
and a test set with 50 observations, i.e. use `sample(1:200, n = 150)` to
randomly choose row indices of the advertising dataset to include in the 
train set. The remaining indices should be used for the test set. Remember
to choose and set the seed for randomization!

```{r}
set.seed(123)
idx <- sample(1:200, size = 150)
train <- sales[idx, ]
test <- sales[-idx, ]

fit <- lm(Sales ~ TV, train)
pred_test <- predict(fit, test)

RMSE <- sqrt(mean((pred_test - test$Sales)^2))

cat("The RMSE for the simple regression model is:", RMSE, ".\n")
```


e. Fit a linear model to the training set, where the sales values are
predicted by the amount of TV advertising. Print the summary of the fitted model.
Then, predict the sales values for the test set and evaluate the test model 
accuracy in terms of root mean squared error (MSE), which measures 
the average level of error between the prediction and the true response.
$$RMSE = \sqrt{\frac{1}{n} \sum\limits_{i = 1}^n(\hat y_i - y_i)^2}$$

f. Fit a multiple linerar regression model including all the variables 'TV',
'radio', 'newspaper' to model the 'sales' in the training set. Then, compute 
the predicted sales for the test set with the new model and evalued the RMSE.  
Did the error decrease from the one correspodning to the previous model?


```{r}
fit.mult <- lm(Sales ~ TV + Radio + Newspaper, train)
pred_test <- predict(fit.mult, test)

RMSE <- sqrt(mean((pred_test - test$Sales)^2))
cat("The RMSE for the multiple regression model is:", RMSE, ".\n")
```

Yes, the RMSE decrease after including more predictors in the model.

g. Look at the summary output for the multiple regression model and note which 
of the coefficient in the model is significant. Are all of them significant?
If not refit the model including only the features found significant.
Which of the models should you choose? 


```{r}
summary(fit.mult)
```

It seems like the 'newspaper' predictor is not significant, so
we will remove it.

```{r}
fit.mult2 <- lm(Sales ~ TV + Radio, train)
pred_test <- predict(fit.mult2, test)

RMSE <- sqrt(mean((pred_test - test$Sales)^2))
cat("The RMSE for the new multiple regression model is:", RMSE, ".\n")
```

The second multiple regression model has a slightly better test error.
We should not use more predictors than necessary, so 'newspaper'
should be discarded from the model.


# Doctor Visits

Data was collected on doctor visits from a sample of 5,190 people in the 1977/1978 Australian Health Survey. Cameron (1986) sought to explain the variation in doctor visits using one or more explanatory variables. The data can be found in an R data set from `library(AER)` accessible with the command `data("DoctorVisits")`. Variable descriptions can be found under `help("DoctorVisits")`

Explore the use of a zero-inflated model for this data. Begin with a histogram of the number of visits, some EDA, and fitting several models. Summarize your results. Compare your results with a standard Poisson.

```{r}
library(AER)
data("DoctorVisits")

ggplot(DoctorVisits) +
  geom_histogram(aes(x=visits))+
  theme_bw()

ggplot(DoctorVisits) +
  geom_histogram(aes(x=visits))+
  facet_grid(gender~private)+
  theme_bw()

df= DoctorVisits %>% group_by(visits) %>% summarise_if(is.numeric, mean)
ggplot(df) +
  geom_point(aes(y=log(visits+1), x=age))+
  theme_bw()

ggplot(df) +
  geom_point(aes(y=log(visits+1), x=income))+
  theme_bw()

```
There are a lot of zeros --- hence the need for a zero-inflated model.

We begin by fitting a standard model and assessing goodness of fit.

```{r}

summary(pois.m1 <- glm(formula = visits ~ ., family = poisson, 
    data = DoctorVisits))

# Goodness-of-fit test
gof.ts = pois.m1$deviance
gof.pvalue = 1 - pchisq(gof.ts, pois.m1$df.residual)
gof.pvalue
```
Now we fit the zero-inflated model:


```{r}
library(pscl)
summary(zi <- zeroinfl(formula = visits ~ health + age + gender | freepoor, data = DoctorVisits))
```

```{r}
vuong(pois.m1,zi)

```
# Movies

Recall the movies data-frame we used ealier in the bootcamp. It contains
information on movies from the last three decates, which was scrapped from
the IMDB database.

```{r}
library(dplyr)
url <- "https://raw.githubusercontent.com/Juanets/movie-stats/master/movies.csv"
movies <- tbl_df(read.csv(url))
```

a. Generate a boxplot of runtimes for action movies and comedies
with jittered points overlaid on top. You might consider setting collor, 
fill and alpha arguments to modify clarity and transparency of the plot.

```{r}
library(tidyverse)
ggplot(movies %>% filter(genre %in% c("Action", "Comedy")),aes(x=genre, y = runtime)) +
  geom_boxplot(outlier.shape = NA) +
  geom_jitter(alpha=0.5, width = 0.1)
```

b. Test a hypothesis that the action movies have higher mean runtime (length)
than the comedies. Is the difference statistically greater than zero
at significance level $\alpha = 0.05$?


```{r}
comedies = movies %>% filter(genre== "Comedy")
actions = movies %>% filter(genre== "Action")
t.test(comedies$runtime, actions$runtime)
```

c. Test the hypothesis that the scores are the same across movie types (keep the movie genre which have at least 20 movies). Plot the data before making a test of your choice. State all the assumptions that you are making when devising your test.

+ The observations are obtained independently and randomly from the population defined by the factor levels
+ The data of each factor level are normally distributed.
+ These normal populations have a common variance. (Levene’s test can be used to check this.)

```{r}

d = movies %>%
  group_by(genre) %>%
  summarise(
    count = n(),
    mean = mean(score, na.rm = TRUE),
    sd = sd(score, na.rm = TRUE)
  ) %>% filter(count>=20)
genres_selected = d$genre
ggplot(movies %>% filter(genre %in% genres_selected), aes(x=genre, y=score, color=genre)) +
  geom_boxplot(outlier.shape = NA) +
  geom_jitter(alpha=0.1, width = 0.1) +
  theme_bw()


res.aov <- aov(score ~ genre, data =movies %>% filter(genre %in% genres_selected))
# Summary of the analysis
summary(res.aov)

```


d. Is there a reason to believe that the scores might differ according to genre? How would you test which one is different (do the test if you have reason to believe that this is the case).

```{r}
TukeyHSD(res.aov)
```

Interpretation:

+ diff: difference between means of the two groups
+ lwr, upr: the lower and the upper end point of the confidence interval at 95% (default)
+ p adj: p-value after adjustment for the multiple comparisons.


e. Now, sppose your friend at Hollywood wants to know the recipe for making a movie with the best margin (gross-budget). Which genre and runtime should he aim for? Could you fit a model to try to help him out?


```{r}
movies$margin = movies$gross - movies$budget
summary(mod <- lm(margin ~ genre+ runtime, data= movies))

```
